Close

%0 Conference Proceedings
%4 sid.inpe.br/sibgrapi/2018/08.27.16.23
%2 sid.inpe.br/sibgrapi/2018/08.27.16.23.42
%@doi 10.1109/SIBGRAPI.2018.00063
%T Unsupervised representation learning using convolutional and stacked auto-encoders: a domain and cross-domain feature space analysis
%D 2018
%A Cavallari, Gabriel B.,
%A Ribeiro, Leonardo S. F.,
%A Ponti, Moacir A.,
%@affiliation USP
%@affiliation USP
%@affiliation USP
%E Ross, Arun,
%E Gastal, Eduardo S. L.,
%E Jorge, Joaquim A.,
%E Queiroz, Ricardo L. de,
%E Minetto, Rodrigo,
%E Sarkar, Sudeep,
%E Papa, João Paulo,
%E Oliveira, Manuel M.,
%E Arbeláez, Pablo,
%E Mery, Domingo,
%E Oliveira, Maria Cristina Ferreira de,
%E Spina, Thiago Vallin,
%E Mendes, Caroline Mazetto,
%E Costa, Henrique Sérgio Gutierrez,
%E Mejail, Marta Estela,
%E Geus, Klaus de,
%E Scheer, Sergio,
%B Conference on Graphics, Patterns and Images, 31 (SIBGRAPI)
%C Foz do Iguaçu, PR, Brazil
%8 29 Oct.-1 Nov. 2018
%I IEEE Computer Society
%J Los Alamitos
%S Proceedings
%K Deep Learning, Representation learning, Feature extraction, Unsupervised feature learning.
%X A feature learning task involves training models that are capable of inferring good representations (transformations of the original space) from input data alone. When working with limited or unlabelled data, and also when multiple visual domains are considered, methods that rely on large annotated datasets, such as Convolutional Neural Networks (CNNs), cannot be employed. In this paper we investigate different auto-encoder (AE) architectures, which require no labels, and explore training strategies to learn representations from images. The models are evaluated considering both the reconstruction error of the images and the feature spaces in terms of their discriminative power. We study the role of dense and convolutional layers on the results, as well as the depth and capacity of the networks, since those are shown to affect both the dimensionality reduction and the capability of generalising for different visual domains. Classification results with AE features were as discriminative as pre-trained CNN features. Our findings can be used as guidelines for the design of unsupervised representation learning methods within and across domains.
%@language en
%3 sibgrapi-2018_Analysis_of_cross_domain_unsupervised_learning.pdf


Close